Brain and Language
○ Elsevier BV
Preprints posted in the last 30 days, ranked by how well they match Brain and Language's content profile, based on 11 papers previously published here. The average preprint has a 0.00% match score for this journal, so anything above that is already an above-average fit.
Kim, J.; Lee, S.; Nam, K.
Show abstract
A central question in psycholinguistics in visual word recognition is whether morphologically complex words are obligatorily decomposed into stems and affixes during visual word recognition or whether whole-word access can occur when forms are frequent and familiar. The present study investigated how morphological complexity and lexical frequency jointly shape neural responses by leveraging Korean nominal inflection, whose transparent stem-suffix structure permits a clean dissociation between base (stem) frequency and surface (whole-word) frequency. Twenty-five native Korean speakers completed a rapid event-related fMRI lexical decision task involving simple and inflected nouns that varied parametrically in both frequency measures. Representational similarity analysis (RSA) revealed robust encoding of surface frequency--but not base frequency--in the inferior frontal gyrus (IFG) pars opercularis and supramarginal gyrus (SMG), with significantly stronger correlations for inflected than simple nouns. Univariate analyses converged with this result: surface frequency selectively increased activation for inflected nouns in inferior parietal regions, whereas base frequency showed no reliable effects in any ROI. These findings challenge models positing obligatory pre-lexical decomposition, instead supporting accounts in which morphological processing is shaped by post-lexical, usage-driven lexical statistics. Taken together, our findings shed light on a distributed perspective on morphological processing, suggesting that structural and statistical factors jointly constrain access to morphologically complex forms.
Vivion, M.; Mathy, F.; Guida, A.; Mondot, L.; Ramanoel, S.
Show abstract
Spatialization in working memory refers to the spatial coding of non-spatial information along a mental horizontal line when encoding verbal material. This phenomenon is thought to support working memory by facilitating order encoding. Although it has been observed for both visually and auditorily presented stimuli, no direct comparison has yet examined whether these modalities rely on similar neural mechanisms. In this study, we investigated whether spatialization in visual and auditory modalities involves shared or distinct patterns of activity within the working-memory network. Forty-nine participants performed both a visual and an auditory working memory SPoARC task of the same verbal material, allowing to study the cortical patterns associated with distinct serial positions at both encoding and recognition across sensory modalities. Whole-brain analyses revealed similar frontoparietal networks across conditions. In addition, a representational similarity analysis (RSA) was conducted to assess the similarity of neural patterns between early and late serial positions in a sequence and across sensory modalities. This multivoxel pattern analysis revealed modality-dependent patterns distinguishing early and late positions in the inferior frontal gyrus. Additional modality-specific effects were observed in the anterior intraparietal sulcus in the visual modality and in the posterior hippocampus in the auditory modality. Drawing on the framework proposed by Bottini & Doeller (2020), we propose that order decoding in the IPS might reflect a low-dimensional spatial coding of order (e.g., along a horizontal axis), whereas order decoding in the hippocampus might reflect higher-dimensional spatial representations or temporal representations.
Allen, S. C.; Koukouvinis, S.; Varjopuro, S. M.; Keitel, A.
Show abstract
Cortical tracking of acoustic features is essential for the neural processing of continuous stimuli such as speech and music. For example, it has been shown that children with dyslexia show atypical cortical tracking. This tracking may therefore reflect a fundamental auditory temporal processing mechanism supporting literacy more generally. In the current pre-registered study, we tested the hypothesis that cortical tracking of speech and music predicts reading ability in healthy young adults (N = 32), evaluated through a lexical decision task. Participants first completed an online session in which they performed a lexical decision task to assess their reading skills. This was followed by an electroencephalography (EEG) session, in which participants listened to a naturalistic short story and a music track. Using mutual information, we showed that neural activity aligned to both speech and music across a wide range of frequencies. Interestingly, cortical tracking was stronger for speech at very low frequencies, while it was stronger for music at higher frequencies. Critically, cortical tracking predicted reaction times in the lexical decision task in a frequency-dependent manner: stronger delta-band tracking (~1-3 Hz) for both speech and music was associated with faster reaction times, whereas stronger alpha-band tracking (~12 Hz) for speech was associated with slower reaction times. These findings remained significant even when controlling for stimulus type, age, musical experience and reading enjoyment. These results suggest that cortical tracking of speech and music reflect a domain-general temporal processing mechanism that is associated with reading ability beyond stimulus-specific features, and beyond development. These findings advance the neurobiological underpinnings of literacy and could potentially be leveraged for developing new reading interventions.
Bahar, N.; Cler, G. J.; Asaridou, S. S.; Smith, H. J.; Willis, H. E.; Healy, M. P.; Chughtai, S.; Haile, M.; Krishnan, S.; Watkins, K. E.
Show abstract
Children with developmental language disorder (DLD) have persistent language learning difficulties and often perform poorly on pseudoword repetition, a task that probes phonological, memory, and speech-motor processes that support vocabulary acquisition. Research on the neural basis of pseudoword repetition in DLD is limited. We used whole-brain functional MRI (fMRI) to examine pseudoword repetition and repetition-based learning in 46 children with DLD (ages 10-15 years) and 71 age-matched children with typical language development. During scanning, children heard and repeated pseudowords paired with visual referents, allowing us to track learning-related changes in neural activity across repetitions. Repeated pseudoword production yielded comparable behavioural learning across groups, with faster productions by later repetitions. Post-scan, form-referent recognition was comparable across groups, whereas pseudoword repetition accuracy was lower in DLD. Pseudoword repetition engaged a distributed neural network, including inferior frontal cortex bilaterally (greater on the left), premotor and sensorimotor cortex, and posterior temporal and occipital regions. Group differences emerged primarily in regions where activity was task negative (i.e., below baseline or deactivated): lateral occipito-parietal cortex (posterior angular gyrus), medial parieto-occipital cortex (retrosplenial), and right posterior cingulate cortex. Learning-related decreases in activity were similar across groups, but region-of-interest analyses showed reduced leftward lateralisation of activity in inferior frontal gyrus in DLD. These findings suggest weaker disengagement of the default mode network during a linguistically demanding task in DLD. Although repetition-based pseudoword learning recruited similar neural mechanisms in both groups, these mechanisms may operate less efficiently in DLD, alongside reduced hemispheric specialisation in inferior frontal cortex. HighlightsO_LISimilar repetition-related neural attenuation across groups during pseudoword learning. C_LIO_LIReduced default-mode network suppression during pseudoword repetition in DLD. C_LIO_LIReduced left-hemisphere specialisation of inferior frontal cortex in DLD. C_LIO_LIRepetition-based learning in DLD supported by less efficient neural networks. C_LI
Lallier, M.; Rius-Manau, C.; 23andMe Research Team, ; Carrion-Castillo, A.
Show abstract
Here, we test the hypothesis that early sustained exposure to complex bilingual environments can positively affect reading development by altering structural interhemispheric connectivity via the corpus callosum (CC). Interhemispheric connectivity has been shown to be inefficient in dyslexia, but also to support compensatory pathways when genetic risk for reading difficulties is present, by enabling the preserved right hemisphere to support a dysfunctional left hemisphere. Mediation models were conducted on children aged 9-10 years (with a 2-year follow-up assessment) from the Adolescent Brain Cognitive Development database (N>10,000). Polygenic scores (PGS) for dyslexia and cognitive performance and continuous bilingualism indices were used as predictors, with reading aloud as the outcome. Bilingualism showed a positive effect on reading partially mediated by the anterior CC, independently of overall brain size. In contrast, genetic predispositions to reading difficulties influenced reading primarily through overall brain size rather than CC connectivity specifically. These two pathways were independent, suggesting that bilingual experience and genetic risk operate through distinct neuroanatomical mechanisms. These findings suggest that recurrent early exposure to complex bilingual environments may shape the brains structural connectivity toward a more balanced and integrated bilateral frontal organisation. The results highlight potential brain compensatory pathways induced by environmental experiences that may support more efficient reading development and mitigate risks for developmental dyslexia.
Saloranta, E.; Tuulari, J. J.; Pulli, E. P.; Audah, H. K.; Barron, A.; Jolly, A.; Rosberg, A.; Mariani Wigley, I. L. C.; Kurila, K.; Yada, A.; Yli-Savola, A.; Savo, S.; Eskola, E.; Fernandes, M.; Korja, R.; Merisaari, H.; Saukko, E.; Kumpulainen, V.; Copeland, A.; Silver, E.; Karlsson, H.; Karlsson, L.; Mainela-Arnold, E.
Show abstract
Previous studies exploring the connection between early language development and brain anatomy have shown that cortical areas relating to individual differences in language skills are diverse and vary depending on the age of child. However, due to lack of large longitudinal samples, current literature is limited in answering the extent to which individual differences in language development prior to school age are reflected in areas of the cortex. To fill this gap, we compared gray matter density between participants that belonged to different longitudinally defined language profiles from 14 months to five years of age in a large population-based sample. Participants were 166 children from the FinnBrain Birth Cohort Study who had longitudinal language data from 14 months to five years of age and magnetic resonance imaging data at five years of age. Three groups of language development were used as per our prior study: persistent low, stable average, and stable high. Voxel-based morphometry metrics were calculated using SPM12 and the three language profile groups were compared to one another. Covariates included sex and age at brain scan. The statistics were thresholded at p < 0.01 and false discovery rate corrected at the cluster level. Of the three longitudinal language profiles, the stable high group had higher gray matter density than the persistent low group in the right superior frontal gyrus. No differences were found between the stable average and stable high groups, nor persistent low and stable average groups. The identified superior frontal cortical area belongs to executive functions neural network. This finding adds to the cumulating evidence that individual differences in language development are reflected in growth of gray matter supporting general processing ability rather than specialized language regions. The results suggest that cognitive development and early language development are linked through shared principles of neural growth, identifiable already at age five. Key pointsO_LIAn association between early language development from 14 months to five years of age and gray matter density differences of the right superior frontal gyrus was found at the age of five years. Children following the strongest language trajectory were more likely to exhibit higher gray matter density of the right superior frontal gyrus than children following the weakest trajectory. C_LIO_LIAs the superior frontal gyrus is part of executive functions network, we propose that individual differences in early language development are more defined by general learning mechanisms supported by those networks, rather than language specific pathways. C_LI
Wang, R.; Guo, Q.; Zeng, X.; Leong, C.; Zhang, C.; Zhang, Y.; Abutalebi, J.; Myachykov, A.
Show abstract
BackgroundThe brains glymphatic system plays a vital role in maintaining neural health. However, little is known about whether second language (L2) immersion can influence this clearance pathway. Methods50 high-proficiency L2 English speakers (mean age: 32.6 years; 78% female) were assessed for glymphatic function using three multimodal MRI markers: BOLD-CSF coupling strength (fMRI), choroid plexus ratio (structural MRI), and DTI-ALPS index (diffusion MRI). Analyses examined relationships between glymphatic markers and L2 immersion duration, age of acquisition (AOA), and active use environment, controlling for age, education, and sex. ResultsL2 immersion duration correlated significantly with better glymphatic function. Longer immersion related to better BOLD-CSF coupling strength (r = -0.315, p < 0.05) and decreased choroid plexus ratios (r = -0.39, p < 0.05), suggesting enhanced brain-CSF coordination and fewer pathological CSF production structures. Mediation analyses demonstrated that immersion influenced ALPS indirectly through effects on choroid plexus morphology and BOLD-CSF coupling. L2 AOA moderated the immersion-coupling relationship: individuals who began learning after age 9.53 showed stronger associations between immersion and BOLD-CSF coupling, though AOA did not moderate choroid plexus effects. As for L2 immersive active is associated with better glymphatic function, while L2 immersive passive and L2 non-immersive active are both unrelated. ConclusionsL2 immersion associates with better glymphatic system function through multiple pathways, including improved brain-CSF coordination, optimized choroid plexus structure, and increased perivascular flow. These findings provide novel neurobiological evidence that bilingual experience may confer neuroprotective benefits through brain waste clearance mechanisms.
Eustace, S. D.; Guediche, S.; Brasiello, L.; Rocha, M.; Correia, J. M.
Show abstract
Speech production requires orchestration of multiple brain systems, including cortical and subcortical areas that support the unfolding of the spoken message across hierarchical linguistic levels, such as phonemes, syllables, words or phrases. Transitions between levels are critical for fluent speech, yet the neural dynamics of, for example, syllable-level and word-level transitions remain unknown. In this electroencephalography (EEG) study, we use time-frequency analysis and source localization to determine differences associated with word-boundary vs. within-word syllable transitions. To this end, pseudoword pairs comprising six consonant-vowel (CV) syllables with different word-boundary positions were used. Fluent human adults produced the utterances at the rhythm of a learned visual metronome (i.e., syllable-by-syllable), such that each syllable was uttered at matching times independently of its relative word position. Accordingly, a target syllable could be either a within-word syllabic transition or a between-word transition, while other linguistic properties, including articulation, stress pattern, co-articulation or prosody, were matched. EEG time-frequency analyses of neural sources successfully revealed sensitivity to hierarchical structure. Neural sources in left and right inferior frontal lobes, as well as left superior temporal lobe were differentially recruited when producing the same exact syllables, in the same exact utterance position, but under different word boundary contexts. A right inferior frontal source showed a robust time-frequency modulation in word transitions that included elevated event-related synchronization in the theta and beta range. Interestingly, despite our efforts to control speech pace across conditions using metronome-based guidance, small, albeit significant timing delays emerged, confirming higher cognitive demands at word boundaries.
Nishio, M.; Ziv, M.; Ellwood-Lowe, M. E.; Ignachi Sanguinetti, J.; Denervaud, S.; Hirsh-Pasek, K.; Golinkoff, R. M.; Mackey, A. P.
Show abstract
Play is a fundamental aspect of childhood and plays a crucial role in the development of creativity, yet its neural mechanisms remain poorly understood. We tested the hypothesis that more frequent play is associated with stronger functional integration among the default mode network (DMN), executive control network (CN), and salience network (SAL), as these cortical networks have been implicated in creativity in adults. In a preregistered study of infants and toddlers (Study 1; N = 143, 10 months-3 years, 67 boys, Baby Connectome Project), parent-reported play and imitation behaviors increased sharply from 1 to 2 years, and were associated with stronger within-DMN connectivity and DMN-CN coupling, controlling for age, sex, and head motion. In middle childhood (Study 2; N = 108, ages 4-11 years, 52 boys), parent-reported play frequency declined with age, as did cross-network coupling involving SAL. However, children who engaged more frequently in play showed higher DMN-SAL and CN-SAL connectivity. Finally, in a quasi-experimental comparison (Study 3; N = 45; ages 4-12 years, 20 boys), children enrolled in a curriculum that includes guided play (Montessori) showed higher DMN-SAL and DMN-CN connectivity than peers in traditional schools, suggesting that pedagogies that center child-led exploration might enable protracted brain network integration. Across these three studies, play was consistently associated with greater integration among DMN, SAL, and CN, a pattern previously linked to creativity in adults. Our findings offer a potential mechanism linking childhood play to later creativity through its role in supporting brain integration during development. Public Significant StatementO_LIPlay is widely believed to nurture childrens creativity, yet the brain mechanisms behind this link are not well understood. C_LIO_LIAcross three studies from infancy to middle childhood, we found that more frequent play was associated with stronger integration among brain networks tied to imagination, attention, and control. C_LIO_LIThese findings suggest that play may help build the neural foundation for later creative thinking. C_LI
Marrazzo, G.; Pimpini, L.; Kochs, S.; De Martino, F.; Valente, G.; Roefs, A.
Show abstract
Despite substantial progress in understanding how visual features of food are processed in the brain, it remains unclear how subjective and nutritional properties, such as perceived palatability, caloric content, and health value, are reflected in neural representational structure. Using functional MRI and representational similarity analysis (RSA), we examined how visual, subjective, and nutritional food properties are encoded in ventral visual cortex. Univariate analyses revealed reliable activation differences between high- and low-calorie foods in lateral occipitotemporal cortex (LOTC) and fusiform gyrus. RSA further revealed a functional dissociation within the ventral stream: LOTC showed systematic correspondence with both visual and subjective dimensions, whereas fusiform cortex exhibited a selective association with perceived caloric content, with both effects persisting after controlling for visual similarity. These results suggest that food-related dimensions not fully captured by the tested visual models are reflected within visual representational spaces, and that LOTC and fusiform cortex show dissociable representational profiles with respect to subjective and perceived nutritional food dimensions.
Madzime, J. S.; Jankiewicz, M.; Meintjes, E. M.; Torre, P.; Laughton, B.; Holmes, M. J.
Show abstract
BackgroundChildren who are HIV-exposed but uninfected (CHEU) face elevated risks of hearing loss and language deficits compared to HIV-unexposed peers. The central auditory system (CAS) undergoes substantial maturational changes during adolescence, yet no neuroimaging study has examined its structural or functional integrity in CHEU. Prior work in this cohort identified white matter (WM) alterations in regions adjacent to the CAS at age 7, and reduced auditory working memory in CHEU relative to unexposed children (CHUU). AimTo characterise WM integrity and functional connectivity (FC) of the CAS and related regions in CHEU at age 11, to investigate structural and functional network topology, and to examine associations between imaging outcomes and neurocognitive function. MethodsForty-eight children aged 11-12 (20 CHEU, 28 CHUU) from an ongoing longitudinal neurodevelopmental cohort underwent 3T MRI including diffusion tensor imaging (DTI) and resting-state fMRI (RS-fMRI). CAS regions (cochlear nucleus/superior olivary complex, inferior colliculus [IC], medial geniculate nucleus [MGN], and primary auditory cortex [PAC]) were manually segmented and combined with an automated atlas. DTI probabilistic tractography was performed, extracting FA, MD, AD, RD, fractional number of tracts, and tract volume. FC was computed using Pearson correlations between regional time series. Graph theory measures (degree, strength, transitivity, nodal and local efficiency) were derived for structural and functional networks. RS-fMRI group comparisons used Bayesian multilevel modelling (matrix-based and region-based analyses), while DTI comparisons used linear models with FDR correction. Neurocognitive testing employed the KABC-II. ResultsNo significant group differences in DTI WM metrics (FA, MD, AD, RD) were observed after FDR correction. CHEU demonstrated higher structural nodal strength in the left IC (FDR-significant) and in the bilateral rostral middle frontal cortex (rMFC) and right cuneus. RS-fMRI revealed lower FC between the bilateral IC in CHEU, alongside reduced FC in the left caudate, left hippocampus CA3, left pericalcarine, and left lingual gyrus. CHEU showed higher FC between the left MGN and right precentral, left postcentral, and right rMFC; the right PAC also showed higher FC to the right rMFC and left postcentral gyrus. No significant group differences were observed in functional nodal measures. No significant associations were found between structural or functional imaging outcomes and neurocognitive scores after multiple comparison correction. DiscussionStructural and functional alterations within the CAS were most prominent in the IC, with increased nodal strength in CHEU potentially reflecting compensatory structural connectivity, and reduced interhemispheric FC between the bilateral IC suggesting disrupted auditory integration. Altered FC between the MGN/PAC and cortical regions, including the rMFC and sensorimotor cortices, may reflect differences in top-down auditory processing. The absence of imaging-cognition associations at age 11 suggests that these connectivity differences do not, at this stage, translate into measurable deficits in auditory or language-related neurocognitive performance. ConclusionThis is the first study to examine functional and structural connectivity of the CAS in CHEU children. HIV exposure is associated with subtle but discernible alterations in IC connectivity and in CAS links to cortical regions at age 11, without detectable neurocognitive correlates. Longitudinal follow-up and inclusion of audiological and ART exposure data are needed to clarify the developmental and functional consequences of these findings.
Yao, J. K.; Mitchell, J.; Davison, A.; Yeatman, J. D.
Show abstract
Individual differences in cognitive abilities have been linked to variability in cortical folding, a stable neuroanatomical scaffold largely established in utero. In the domain of reading, recent findings in small groups of typical readers suggest that a sulcal interruption (superficial annectant gyrus, gyral gap) in the left posterior occipital temporal sulcus (lhpOTS) predicts better reading skills, posing the lhpOTS as a potential early biomarker of reading difficulties. However, it remains unknown whether this relationship found in typical readers generalizes to the dyslexic population and whether the lhpOTS can serve as a biomarker for dyslexia or predict response to targeted instruction.To fill these gaps, we examine the patterns of the lhpOTS in 209 children, including children with dyslexia, from four independently-collected samples. In typical readers, we find that the relationship between the lhpOTS and reading skills is robust, replicating across binary and continuous quantifications of the sulcal interruption. However, lhpOTS patterns neither distinguish dyslexic children from typical readers nor do they predict response to intervention. Instead, targeted reading intervention drives long-term gains in reading skills that are equivalent irrespective of VOTC anatomy. Together, these findings distinguish neuroanatomical correlates of skilled reading from determinants of reading impairment and learning capacity and emphasize the importance of the educational environment in supporting reading acquisition for children with dyslexia. SIGNIFICANCE STATEMENTEarly predictors of dyslexia are important for understanding the etiology of reading difficulties and informing early intervention. One candidate biomarker for dyslexia is the left posterior occipital temporal sulcus (lhpOTS), a neuroanatomical feature established before birth. In typical readers, the presence of an interruption in the lhpOTS has been linked to better reading skills. Here, we examine this neuroanatomical feature in 209 children with and without dyslexia. While the lhpOTS reliably relates to reading skill in typical readers, it neither differentiates dyslexic from typical readers nor predicts response to intensive reading intervention. These results show that brain anatomy reflects reading proficiency but does not constrain learning and highlights the power of targeted intervention to support reading development.
Bair, M. B.; Long, N. M.
Show abstract
It is critical to identify which factors induce specific brain states as these large-scale patterns of coordinated neural activity drive downstream processing and behavior. The retrieval state, a brain state engaged when attempting to retrieve the past, is thought to specifically support episodic memory, remembering experiences within a spatiotemporal context, as opposed to semantic memory, remembering general knowledge. However, we hypothesize that the retrieval state reflects internal attention engaged to access stored episodic and semantic information. To test these alternatives, we recorded scalp electroencephalography while participants made episodic, semantic, or perceptual judgments, and applied an independently validated mnemonic state classifier to measure retrieval state engagement. We found that retrieval state engagement was greater for both episodic and semantic judgments compared to perceptual judgments. These findings suggest that the retrieval state reflects a domain-general internal attention process that supports not just episodic memory, but internally directed cognition.
Figarola, V.; Liang, W.; Luthra, S.; Parker, E.; Winn, M.; Brown, C.; Shinn-Cunningham, B. G.
Show abstract
Listeners face many challenges when trying to maintain attention to a target source in everyday settings; for instance, reverberation distorts acoustic cues and interruptions capture attention. However, little is known about how these challenges affect the ability to maintain selective attention. Here, we measured syllable recall accuracy and pupil dilation during a spatial selective attention task that was sometimes disrupted. Participants heard two competing, temporally interleaved syllable streams presented in pseudo-anechoic or reverberant environments. On randomly selected trials, a sudden interruption occurred mid-sequence. Compared to anechoic trials, reverberant performance was worse overall, and the interrupter disrupted performance. In uninterrupted trials, reverberation reduced peak pupil dilation both when it was consistent across all stimuli in a block and when it was randomized trial to trial, suggesting temporal smearing reduced clarity of the scene and the salience of events in the ongoing streams. Pupil dilations in response to interruptions indicated perceptual salience was strong across reverberant and anechoic conditions. Specifically, baseline pupil size before trials did not vary across room conditions, and mixing or blocking of trials (altering stimulus expectations) had no impact on pupillary responses. Together, these findings highlight that stimulus salience drives cognitive load more strongly than does task performance.
Chen, W.; Pell, M.; Jiang, X.
Show abstract
People encounter AI voices daily. Existing behavioral studies suggest listeners rely on prosodic cues such as intonation and expressiveness to detect audio deepfakes, reporting that AI voices sound prosodically less rich than human voices. To test whether prosodic processing drives deepfake discrimination in the neural time course of voice processing, we recorded electroencephalographic (EEG) data while participants listened to human and AI-generated speakers producing utterances in confident vs. doubtful prosody (tone of voice), with attention directed toward memorizing speaker names. We used voice cloning to control for speaker identity confounds between human and AI voices. Multivariate pattern analysis revealed that neural discrimination of human vs. AI voices emerged rapidly regardless of prosody (confident: 176 ms; doubtful: 134 ms), substantially preceding prosody discrimination (confident vs. doubtful within human voices: 2066 ms; within AI voices: 1366 ms). Acoustic analysis confirmed that prosodic distinctions became classifiable only at utterance offset (90% normalized duration), converging with neural evidence that prosody requires near-complete temporal integration. This temporal dissociation between rapid voice source discrimination and late-emerging prosody decoding suggests that prosody plays a smaller role in audio deepfake detection than listeners retrospectively report. Representational similarity analysis further revealed that spectral envelope features (mel-frequency cepstral coefficients; MFCC), rather than the visually salient high-frequency energy differences, drove neural human-AI discrimination, with MFCCs earliest independent contribution (228 ms) closely following the MVPA decoding onset (134-176 ms). Future studies may manipulate specific acoustic components to establish the causal sources of this rapid and sustained neural discrimination. Significance StatementPeople encounter AI voices daily, in phone calls, navigation apps, supermarket checkouts, and subway announcements. Using electroencephalography, we show that the human brain automatically and rapidly distinguishes everyday AI voices from human speech, even without conscious attention to voice source. Although people may attribute this ability to AI voices sounding monotone or prosodically unnatural, the brain relies on subtler acoustic signatures, enabling discrimination before prosodic information becomes available. Attempts to identify the specific acoustic features driving this neural detection were inconclusive, pointing to the need for future causal investigations. We encourage engineers and policymakers to ensure AI voices remain perceptually detectable, as increasingly humanlike AI voices could cognitively disadvantage the general public if they become indistinguishable from human speech.
Abenes, I.; Jiahui, G.
Show abstract
Functional localizer scans have long served as the classic method for mapping individualized functional topographies, but they require dedicated scan time and can be difficult to implement in neuropsychological populations. Previous work has shown that individualized functional topographies can be estimated with high fidelity in typical participants using hyperalignment, but it remains unknown whether this approach generalizes to populations with functional deficits. Here, we tested this question in developmental prosopagnosia (DP), a neuropsychological condition characterized by severe face recognition impairments. Using two independent datasets that included both DP and control participants, we estimated individualized category-selective functional topographies from independent participants using hyperalignment derived from either a task-based scan or a naturalistic movie-viewing scan. Across datasets, whole-brain correlations and searchlight analyses showed that predicted topographies were highly similar to topographies estimated from participants own localizer data, especially in cortical areas with strong category-selective responses. Hyperalignment successfully recovered idiosyncratic features of category-selective topographies and consistently outperformed anatomical alignment. Importantly, predictions generalized across groups, such that individualized topographies in DPs could be estimated from control participants and vice versa. In addition, predicted topographies preserved the reduced face selectivity in DPs previously reported in the literature. These findings support a hyperalignment-based framework for estimating individualized functional topographies in neuropsychological populations without requiring separate localizer scans, and provide a foundation for integrating existing datasets to study the underlying neural basis in DP and other atypical populations.
Augsten, M.-L.; Lindenbeck, M. J.; Laback, B.
Show abstract
Cochlear implant (CI) users typically experience difficulties perceiving musical harmony due to a restricted spectro-temporal resolution at the electrode-nerve interface, resulting in limited pitch perception. We investigated how stimulus parameters affect discrimination of complex-tone triads (three-voice chords), aiming to identify conditions that maximize perceptual sensitivity. Six post-lingually deafened CI listeners completed a same/different task with harmonic complex tones, while spectral complexity, voice(s) containing a pitch change, and temporal synchrony (simultaneous vs. sequential triad presentation) were manipulated. CI listeners discriminated harmonically relevant one-semitone pitch changes within triads when spectral complexity was reduced to three or five components per voice, with significantly better performance for three-component compared to nine-component tones. Sensitivity was observed for pitch changes in the high voice or in both high and low voices, but not for changes in only the low voice. Single-voice sensitivity predicted simultaneous-triad sensitivity when controlling for spectral complexity and voice with pitch change. Contrary to expectations, sequential triad presentation did not improve discrimination. An analysis of processor pulse patterns suggests that difference-frequency cues encoded in the temporal envelope rather than place-of-excitation cues underlie perceptual triad sensitivity. These findings support reducing spectral complexity to enhance chord discrimination for CI users based on temporal cues.
Westner, B. U.; Luo, Y.; Piai, V.
Show abstract
Both episodic memory and word retrieval have been linked to power decreases in the alpha and beta oscillatory bands, but these patterns have rarely been related to each other, partly due to a lack of methodological approaches available. In this explorative study, we investigate the similarities and dissimilarities in the oscillatory fingerprints of the retrieval of words and episodes by directly comparing the activity patterns across time, frequency, and space. We acquired electroencephalography (EEG) data of participants performing a language and an episodic memory task based on the same stimulus material. With a newly developed approach, we directly compared the source-reconstructed oscillatory activity using mutual information and a feature-impact analysis. While left temporal and frontal regions showed dissimilarities between the tasks, right-hemispheric parietal regions exhibited similarities. We speculate that this could indicate a homologous function of these regions, potentially sharing less-specific representations between the tasks. We further uncovered a dissociation of the alpha and beta bands regarding the similarity across tasks. While the beta band was dissimilar between word and episodic memory retrieval, the alpha band seemed to contribute to the similarity we observed in right parietal regions. Whether this points to a task-unspecific function of the alpha band or a functional role in the retrieval process of the presumed representations, remains to be determined. In summary, we present an approach to study similarity across tasks using the temporal, spectral, and spatial dimensions of EEG data, and present results of exploring the shared oscillatory fingerprints between episodic memory and word retrieval.
Ahamdi, S. S.; Fridriksson, J.; Den Ouden, D.
Show abstract
Language impairments in aphasia are characterized by various representational disruptions that may be reflected in discourse production. This research examines the capacity of transformer-based language models, particularly GPT-2, to serve as a computational framework for analyzing variations in aphasic narrative speech. A longitudinal dataset of narrative speech samples collected at six time points from individuals with aphasia (N = 47) was utilized as part of an intervention study. All transcripts were processed via the GPT-2 language model to obtain activation values from each of the 12 transformer layers. Statistically significant differences in activation magnitude across aphasia subtypes were found at every layer (all p < .001), with the most pronounced effects in the deeper layers. Pairwise Tukey HSD tests revealed consistent distinctions between Brocas aphasia and both Anomic and Wernickes aphasia, suggesting a shared activation profile between the latter two. Longitudinal tests revealed significant changes over time, especially in the final three layers (10-12). These findings suggest that transformer-based activation patterns reflect meaningful variation in aphasic discourse and could complement current diagnostic tools. Overall, GPT-2 provides a scalable tool to model representational dynamics in aphasia and enhance the clinical interpretability of deep language models.
King, C. D.; Zhu, T.; Groh, J. M.
Show abstract
Information about eye movements is necessary for linking auditory and visual information across space. Recent work has suggested that such signals are incorporated into processing at the level of the ear itself (Gruters, Murphy et al. 2018). Here we report confirmation that the eye movement signals that reach the ear can produce perceptual consequences, via a case report of an unusual participant with tensor tympani myoclonus who hears sounds when she moves her eyes. The sounds she hears could be recorded with a microphone in the ear in which she hears them (left), and occurred for large leftward eye movements to extreme orbital positions of the eyes. The sounds elicited by this participants eye movements were reminiscent of eye movement-related eardrum oscillations (EMREOs, (Gruters, Murphy et al. 2018, Brohl and Kayser 2023, King, Lovich et al. 2023, Lovich, King et al. 2023, Lovich, King et al. 2023, Abbasi, King et al. 2025, Sotero Silva, Kayser et al. 2025, King and Groh 2026, Leon, Ramos et al. 2026, Sotero Silva, Brohl et al. 2026)), but were larger and longer lasting than classical EMREOs, helping to explain why they were audible to her. Overall, the observations from this patient help establish that (a) eye movement-related signals specifically reach the tensor tympani muscle and that (b) when there is an abnormality involving that muscle, such signals can lead to actual audible percepts. Given that the tensor tympani contributes to the regulation of sound transmission in the middle ear, these findings support that eye movement signals reaching the ear have functional consequences for auditory perception. The findings also expand the types of medical conditions that produce gaze-evoked tinnitus, to date most commonly observed in connection with acoustic neuromas.